Start

  • The numeric files has few additional features, but there is no description of the features so I will just go ahead with the file with both numeric and categorical features

EDA

Note: I have only done the EDA to answer the asked questions. I have not done any EDA for the purpose of feature engineering or feature selection.

Missingness

rm(list = ls())
source("utils.R")

So, no missing data. Yayyy!

Credit Worthiness

Before going into exploring relationship of predictors with the target, let’s first clearly define the target

Credit worthiness for a group of observations can be measured by Good/Total proportion. Higher the proportion, higher the credit worthiness

Credit History

Question: Would a person with critical credit history, be more credit worthy?

Again, let’s first define what critical means. In the absence of any concrete definition, I will assume ‘critical’ roughly means more existing credits i.e. it increase from A30 to A35

load("data_intermediate/data_cleaning_1_output.RData")

Critical has positive association with credit worthiness

Age

Q. Are young people more creditworthy?

sapply(data_mix, function(x) sum(is.na(x)))
         checking_account_status               duration_in_months                   credit_history 
                               0                                0                                0 
                         purpose                    credit_amount           savings_account_status 
                               0                                0                                0 
        present_employment_since installment_as_percent_of_income                 marital_sex_type 
                               0                                0                                0 
           role_in_other_credits           present_resident_since                      assset_type 
                               0                                0                                0 
                             age          other_installment_plans                     housing_type 
                               0                                0                                0 
          count_existing_credits                  employment_type                 count_dependents 
                               0                                0                                0 
                   has_telephone                is_foreign_worker                 is_credit_worthy 
                               0                                0                                0 

The distributions are quite overlapping. But there are more young in “Bad” compared to “Good”, and that is also visible in the difference in means. > So, young people seem slightly less credit worthy.

But let’s break the age into groups to see finer details

plot_list <- plot_target_cat_feature_cont(data_mix, "age", "is_credit_worthy")
`summarise()` ungrouping output (override with `.groups` argument)
ggplotly(
  plot_list[[1]]
)

“Bad” is quite low for the (34, 39] age group

Credit Accounts

Q. Would a person with more credit accounts, be more credit worthy?

I am assuming more credit accounts is same as “Number of existing credits at this bank” i.e. ‘count_existing_credits’

plot_list = plot_target_cat_feature_cat(data_mix, "age_groups", "is_credit_worthy", "Good", "Bad")
`summarise()` ungrouping output (override with `.groups` argument)
Ignoring unknown aesthetics: text
ggplotly(
  plot_list[[1]]
)
plot_list <- plot_target_cat_feature_cat(data_mix, "count_existing_credits", "is_credit_worthy", "Good", "Bad")
`summarise()` ungrouping output (override with `.groups` argument)
Ignoring unknown aesthetics: text

Data is too unreliable to say anything on the relationship between no. of credit accounts and credit worthiness

Feature Engineering & Selection

As mentioned earlier I didn’t do any EDA from featre engineering perspective. So, there is no feature engineering.

For feature selection I have used Boruta, which I have found to be the best feature selection technique almost always. Below is how the Boruta plot looks like:

ggplotly(
  plot_list[[1]]
)

Selected features are:

ggplotly(
  plot_list[[2]]
  , tooltip = "text"
)

Modeling

Strategy

It is worse to class a customer as ‘Good’ when they are ‘Bad’, than it is to class a customer as bad when they are good.

Let ‘Good’ be the positive class, and ‘Bad’ be the negative class. So the above statement will translate to:

False Positives (FPs) are more expensive than False Negatives (FNs)

Such cases fall under **Cost Sensitive Learning" strategy, and followong sub-strategies can be followed decided under it:

Strategy Options

  • Modeling Strategies for cost sensitive learning
    • Change cost function
      • Change the function itself
        • the main function
        • penalty component
      • Change function parameters
        • oversample positive class
          • synthetic sample generation (like SMOTE)
          • give more weight
        • undersample sample negative class
          • give less weight
    • Optimize thresholds that are used for converting output probabilities into class labels - valid only for models which output probabilities
    • Ensembling
  • Evaluation Strategies for Cost sensitive classification
    • Favour Precision over Accuracy or Recall
    • Give weights to different buckets in confusion matrix, and use that to construct a custom evaluation metric

Options that I will explore

Models

I will try the following three models: - Logistic Regression - Boosted Trees: GBM - Random Forest

Modeling Strategy

  • Will optimize thresholds for all the models
  • give more weight to positive class, I will tune the weighing parameter: will do this only for GBM, just to showcase

Evaluation Strategy

I will go with a Custom evaluation metric:

I have assigned follwing weights to different buckets of the confusion matrix to penalize each bucket differently

load("data_intermediate/data_after_fs.RData")
colnames(data_after_fs)
 [1] "checking_account_status"          "duration_in_months"              
 [3] "credit_history"                   "purpose"                         
 [5] "credit_amount"                    "savings_account_status"          
 [7] "present_employment_since"         "installment_as_percent_of_income"
 [9] "role_in_other_credits"            "assset_type"                     
[11] "age"                              "other_installment_plans"         
[13] "housing_type"                     "employment_type"                 
[15] "is_credit_worthy"                

There is no particular reason for these values, just their relative differences are important because they penalize FPs more than FNs. PLus, I am rewarding TPs (True Positives)

Now, the custom metric is just the normalized sum-product of these weights and the confusion matrix of the model. Let’s call it “credit_cost”.

Splitting

I have 80:20 splitting. For validation, I will be using cross-validation wherever required.

Baseline

load("data_intermediate/costs.RData")

I am taking baseline as predicting everybody as "Good’

Train credit_cost

print(costs)
          Reference
Prediction Good Bad
      Good -0.4   1
      Bad   0.2   0

Test credit_cost

load("data_intermediate/splitted_data.RData")

Logistic Regression

Train Results:

model_glm_base <- h2o::h2o.loadModel("models/GLM_model_R_1595304405106_1")

Boosted Trees - GBM

Train Results:

model_gbm_best <- h2o.loadModel("models/gbm_grid_11_model_3")

Random Forest

Train Results:

h2o.loadGrid(
  "models/drf_grid_11"
)
H2O Grid Details
================

Grid ID: drf_grid_11 
Used hyper parameters: 
  -  col_sample_rate_per_tree 
  -  max_depth 
  -  mtries 
  -  ntrees 
  -  sample_rate 
Number of models: 7 
Number of failed models: 0 

Hyper-Parameter Search Summary: ordered by increasing logloss
drf_gridperf_2_byPrecision <- h2o.getGrid(grid_id = "drf_grid_11",
                                          sort_by = "precision",
                                          decreasing = TRUE)

drf_gridperf_2_byAUC <- h2o.getGrid(grid_id = "drf_grid_11",
                                    sort_by = "auc",
                                    decreasing = TRUE)

# Grab the top GBM model, chosen by validation AUC
model_drf_best <- h2o.getModel(drf_gridperf_2_byAUC@model_ids[[1]])

Comparison

train_pred_df_drf_base <- get_prediction(model_drf_best, train_h2o)

  |                                                                                                        
  |                                                                                                  |   0%
  |                                                                                                        
  |==================================================================================================| 100%
train_results_drf <- get_results(train_pred_df_drf_base, costs)
opt_threshold_drf = optimize(get_cost_given_threshold, c(0.1, 0.9), truth_pred_df=train_pred_df_drf_base, costs=costs)

test_pred_df_drf_base_default <- get_prediction(model_drf_best, test_h2o)

  |                                                                                                        
  |                                                                                                  |   0%
  |                                                                                                        
  |==================================================================================================| 100%
test_pred_df_drf_base <- get_prediction_class_given_threshold(test_pred_df_drf_base_default, opt_threshold_drf$minimum)
test_results_drf <- get_results(test_pred_df_drf_base, costs)

Credit_cost and Pricision are in sync.

train results are best for GBM. But its overfitting, i.e. variance is high, so not that great results on test.

test results are best for Random Forest. It has less variance then GBM, but bias is higher.

It may seem like that GBM is a better model, but we still haven’t seen the uncertainity (variance) in the results. Difference between train and test set results give some idea about it, but its better to see it on cross-validated results.

train_results_drf$results
Confusion Matrix and Statistics

          Reference
Prediction Good Bad
      Good  535  52
      Bad    26 189
                                              
               Accuracy : 0.903               
                 95% CI : (0.88, 0.922)       
    No Information Rate : 0.7                 
    P-Value [Acc > NIR] : < 0.0000000000000002
                                              
                  Kappa : 0.761               
                                              
 Mcnemar's Test P-Value : 0.00464             
                                              
            Sensitivity : 0.954               
            Specificity : 0.784               
         Pos Pred Value : 0.911               
         Neg Pred Value : 0.879               
             Prevalence : 0.700               
         Detection Rate : 0.667               
   Detection Prevalence : 0.732               
      Balanced Accuracy : 0.869               
                                              
       'Positive' Class : Good                
                                              
model_comparison_df <- data.frame(
  models = c("baseline", "Logistic Regression", "GBM", "Random Forest"),
  train_credit_cost = c(baseline_train_cost, opt_threshold_glm$objective, opt_threshold_gbm$objective, opt_threshold_drf$objective),
  train_precision = c(baseline_train_precision, train_results_glm$results$byClass["Pos Pred Value"], train_results_gbm$results$byClass["Pos Pred Value"], train_results_drf$results$byClass["Pos Pred Value"]),
  test_credit_cost = c(baseline_test_cost, test_results_glm$cost, test_results_gbm$cost, test_results_drf$cost),
  test_precision = c(baseline_test_precision, test_results_glm$results$byClass["Pos Pred Value"], test_results_gbm$results$byClass["Pos Pred Value"], test_results_drf$results$byClass["Pos Pred Value"])
)

model_comparison_df

Not much difference here too, DRF seems only slightly better but that may change with fold assignment. For GBM, I did positive class upsample tuning but didn’t tune other hyperparameters. And for DRF I did the exact opposite. So, both the models have a lot of scope of tuning, and I am not at a stage to pick the right model

Important Features

We can see feature importance of either GBM or DRF, but DRF gives a better plot without breaking categorical features into its classes, so we will use DRF.

temp <- model_gbm_best@model$cross_validation_metrics_summary
temp_1 <- tibble(metrics = row.names(temp)) %>% 
  bind_cols(temp %>% as_tibble())
datatable(temp_1[,1:3], rownames = FALSE, filter="none", options = list(scrollX=T, scrollY=T, pageLength = 5, scrollCollapse=FALSE, paging=FALSE))

Topp-3 features are “checking_account_status”, “duration_in_months”, and “credit_amount”

Profiling of best credit-worthy person

To profile a ‘Good’ credit worthy person as per the model, let’s explore the relationship of top predictors with the predicted class for the DRF model.

h2o.varimp_plot(model_drf_best)

train_pred_df_drf_default <- get_prediction(model_drf_best, train_h2o)

  |                                                                                                        
  |                                                                                                  |   0%
  |                                                                                                        
  |==================================================================================================| 100%
train_pred_df_drf_opt <- get_prediction_class_given_threshold(train_pred_df_drf_default, opt_threshold_drf$minimum)

test_pred_df_drf_default <- get_prediction(model_drf_best, test_h2o)

  |                                                                                                        
  |                                                                                                  |   0%
  |                                                                                                        
  |==================================================================================================| 100%
test_pred_df_drf_opt <- get_prediction_class_given_threshold(test_pred_df_drf_default, opt_threshold_drf$minimum)

all_pred_df_drf_opt <- bind_rows(train_pred_df_drf_opt, test_pred_df_drf_opt)
plot_list = plot_target_cat_feature_cat(all_pred_df_drf_opt, "checking_account_status", "is_credit_worthy_pred_class", "Good", "Bad")
`summarise()` ungrouping output (override with `.groups` argument)
The `x` argument of `as_tibble.matrix()` must have unique column names if `.name_repair` is omitted as of tibble 2.0.0.
Using compatibility `.name_repair`.
This warning is displayed once every 8 hours.
Call `lifecycle::last_warnings()` to see where this warning was generated.Ignoring unknown aesthetics: text
ggplotly(
  plot_list[[1]]
)
plot_list = plot_target_cat_feature_cont(all_pred_df_drf_opt, "duration_in_months", "is_credit_worthy_pred_class")
`summarise()` ungrouping output (override with `.groups` argument)
ggplotly(
  plot_list[[1]]
)
plot_list = plot_target_cat_feature_cont(all_pred_df_drf_opt, "credit_amount", "is_credit_worthy_pred_class")
`summarise()` ungrouping output (override with `.groups` argument)
ggplotly(
  plot_list[[1]]
)

So, the best credit worthy person would have a following profile:
- checking_account_status is “A14” i.e. no checking account
- duration_in_months is less than 12 month i.e. a year
- credit_amount is less than 2k
- credit_history is “A34” i.e. critical account/other existing credits
- Purpose is A43 i.e. radio/television

This seems slightly unintuitive, but I will have to go into model explainibility to get better insights, and currently the time is short for that

Things to do in future

  • EDA driven Feature Engineering
  • EDA driven Feature Selection
  • Better tuning
    • with cross validation on credit_cost or Precision
    • Bayesian Optimiation
  • Model Explainibility
---
title: "epiFi Assignment"
author: "Sahil Maheshwari"
date: "`r Sys.Date()`"
output:
  html_document: 
    df_print: paged
    toc: yes
    toc_depth: 4
    toc_float:
      collapsed: yes
      smooth_scroll: yes
  html_notebook:
    df_print: paged
    code_folding: "hide"
    theme: united
    toc: yes
    toc_depth: 4
    toc_float:
      collapsed: yes
      smooth_scroll: yes
editor_options:
  chunk_output_type: inline
---

```{r, setup, include=FALSE}
options("scipen"=100, "digits"=4, "knitr.table.format"="pandoc")
```

```{r, global_options, include=FALSE}
knitr::opts_chunk$set(echo = FALSE, warning = FALSE, error = FALSE, message = TRUE,
                      cache = FALSE, cache.lazy = FALSE,
                      tidy = TRUE, highlight = TRUE, collapse = TRUE,
                      fig.fullwidth=TRUE, fig.align = "center", fig.width = 10)
```

```{r, include=FALSE}
rm(list = ls())
source("utils.R")
```

# Start  
- The numeric files has few additional features, but there is no description of the features so I will just go ahead with the file with both numeric and categorical features


# EDA
Note: I have only done the EDA to answer the asked questions. I have not done any EDA for the purpose of feature engineering or feature selection.  

```{r, include=FALSE}
load("data_intermediate/data_cleaning_1_output.RData")
```

## Missingness
```{r}
sapply(data_mix, function(x) sum(is.na(x)))
```

So, no missing data. Yayyy!

## Credit Worthiness
Before going into exploring relationship of predictors with the target, let's first clearly define the target

Credit worthiness for a group of observations can be measured by Good/Total proportion. Higher the proportion, higher the credit worthiness

## Credit History
> Question: Would a person with critical credit history, be more credit worthy?

Again, let's first define what critical means. In the absence of any concrete definition, I will assume 'critical' roughly means more existing credits i.e. it increase from A30 to A35

```{r, message=FALSE}
plot_list = plot_target_cat_feature_cat(data_mix, "credit_history", "is_credit_worthy", "Good", "Bad")

ggplotly(
  plot_list[[1]]
)
```

> Critical has positive association with credit worthiness

## Age
> Q. Are young people more creditworthy?  

```{r, message=FALSE}
plot_list <- plot_target_cat_feature_cont(data_mix, "age", "is_credit_worthy")

ggplotly(
  plot_list[[1]]
)
```

The distributions are quite overlapping. But there are more young in "Bad" compared to "Good", and that is also visible in the difference in means. 
> So, young people seem slightly less credit worthy.

But let's break the age into groups to see finer details

```{r, include=FALSE}
data_mix$age_groups <- cut(data_mix$age, breaks = c(18, 24, 29, 34, 39, 49, 64, Inf), ordered_result = T)
```

```{r, message=FALSE}
plot_list = plot_target_cat_feature_cat(data_mix, "age_groups", "is_credit_worthy", "Good", "Bad")

ggplotly(
  plot_list[[1]]
)
```

> "Bad" is quite low for the (34, 39] age group

## Credit Accounts
> Q. Would a person with more credit accounts, be more credit worthy?

I am assuming more credit accounts is same as "Number of existing credits at this bank" i.e. 'count_existing_credits' 

```{r, include=FALSE}
plot_list <- plot_target_cat_feature_cat(data_mix, "count_existing_credits", "is_credit_worthy", "Good", "Bad")
```

```{r, message=FALSE}
ggplotly(
  plot_list[[1]]
)
```

```{r}
ggplotly(
  plot_list[[2]]
  , tooltip = "text"
)
```

> Data is too unreliable to say anything on the relationship between no. of credit accounts and credit worthiness

# Feature Engineering & Selection  
As mentioned earlier I didn't do any EDA from featre engineering perspective. So, there is no feature engineering.  

For feature selection I have used Boruta, which I have found to be the best feature selection technique almost always. Below is how the Boruta plot looks like:
```{r}
load("models/boruta_train_obj.RData")

plot(boruta_train_obj, xlab = "", xaxt = "n")
lz<-lapply(1:ncol(boruta_train_obj$ImpHistory),function(i)
  boruta_train_obj$ImpHistory[is.finite(boruta_train_obj$ImpHistory[,i]),i])
names(lz) <- colnames(boruta_train_obj$ImpHistory)
Labels <- sort(sapply(lz,median))
axis(side = 1,las=2,labels = names(Labels),
     at = 1:ncol(boruta_train_obj$ImpHistory), cex.axis = 0.7)
```

Selected features are:
```{r}
load("data_intermediate/data_after_fs.RData")
colnames(data_after_fs)
```


# Modeling
## Strategy
> It is worse to class a customer as 'Good' when they are 'Bad', than it is to class a customer as bad when they are good.  

Let 'Good' be the positive class, and 'Bad' be the negative class. So the above statement will translate to:   

> False Positives (FPs) are more expensive than False Negatives (FNs) 

Such cases fall under **Cost Sensitive Learning" strategy, and followong sub-strategies can be followed decided under it:

### Strategy Options
- Modeling Strategies for cost sensitive learning
  - Change cost function
    - Change the function itself
      - the main function
      - penalty component
    - Change function parameters
      - oversample positive class
        - synthetic sample generation (like SMOTE)
        - give more weight
      - undersample sample negative class
        - give less weight
  - Optimize thresholds that are used for converting output probabilities into class labels - valid only for models which output probabilities
  - Ensembling
  
- Evaluation Strategies for Cost sensitive classification
  - Favour Precision over Accuracy or Recall
  - Give weights to different buckets in confusion matrix, and use that to construct a custom evaluation metric

### Options that I will explore
### Models
I will try the following three models:
- Logistic Regression
- Boosted Trees: GBM
- Random Forest

### Modeling Strategy
- Will optimize thresholds for all the models
- give more weight to positive class, I will tune the weighing parameter: will do this only for GBM, just to showcase
  
### Evaluation Strategy
I will go with a Custom evaluation metric:
  
```{r, include=FALSE}
load("data_intermediate/costs.RData")
```

I have assigned follwing weights to different buckets of the confusion matrix to penalize each bucket differently
```{r}
print(costs)
```

There is no particular reason for these values, just their relative differences are important because they penalize FPs more than FNs. PLus, I am rewarding TPs (True Positives)

Now, the custom metric is just the normalized sum-product of these weights and the confusion matrix of the model. Let's call it "credit_cost".

## Splitting
I have 80:20 splitting. For validation, I will be using cross-validation wherever required.

## Baseline
```{r}
load("data_intermediate/splitted_data.RData")
```

I am taking baseline as predicting everybody as "Good'

Train credit_cost
```{r}
train_truth_table = table(data_train$is_credit_worthy)

# give loan to everybody
baseline_train_cost = as.numeric((train_truth_table['Good'] * costs[1, 1] + train_truth_table['Bad'] * costs[1, 2]) / nrow(data_train))
baseline_train_precision = train_truth_table['Good']/nrow(data_train)
message("Baseline Train Cost: ", baseline_train_cost, "\nBaseline Train Precision: ", baseline_train_precision)
```

Test credit_cost
```{r}
test_truth_table = table(data_test$is_credit_worthy)

# give loan to everybody
baseline_test_cost = as.numeric((test_truth_table['Good'] * costs[1, 1] + test_truth_table['Bad'] * costs[1, 2]) / nrow(data_test))
baseline_test_precision = test_truth_table['Good']/nrow(data_test)
message("Baseline Test Cost: ", baseline_test_cost, "\nBaseline Test Precision: ", baseline_test_precision)
```


## Logistic Regression
```{r, include=FALSE}
h2o.init()

train_h2o <- as.h2o(data_train)
test_h2o  <- as.h2o(data_test)

y <- "is_credit_worthy"
x <- setdiff(names(train_h2o), y)
```

```{r, include=FALSE}
model_glm_base <- h2o::h2o.loadModel("models/GLM_model_R_1595304405106_1")
```

```{r, include=FALSE}
train_pred_df_glm_base <- get_prediction(model_glm_base, train_h2o)
train_results_glm <- get_results(train_pred_df_glm_base, costs)
opt_threshold_glm = optimize(get_cost_given_threshold, c(0.1, 0.9), truth_pred_df=train_pred_df_glm_base, costs=costs)

test_pred_df_glm_base_default <- get_prediction(model_glm_base, test_h2o)
test_pred_df_glm_base <- get_prediction_class_given_threshold(test_pred_df_glm_base_default, opt_threshold_glm$minimum)
test_results_glm <- get_results(test_pred_df_glm_base, costs)
```

Train Results:  
```{r}
train_results_glm$results
```


## Boosted Trees - GBM
```{r, include=FALSE}
model_gbm_best <- h2o.loadModel("models/gbm_grid_11_model_3")
```

```{r, include=FALSE}
train_pred_df_gbm_base <- get_prediction(model_gbm_best, train_h2o)
train_results_gbm <- get_results(train_pred_df_gbm_base, costs)
opt_threshold_gbm = optimize(get_cost_given_threshold, c(0.1, 0.9), truth_pred_df=train_pred_df_gbm_base, costs=costs)

test_pred_df_gbm_base_default <- get_prediction(model_gbm_best, test_h2o)
test_pred_df_gbm_base <- get_prediction_class_given_threshold(test_pred_df_gbm_base_default, opt_threshold_gbm$minimum)
test_results_gbm <- get_results(test_pred_df_gbm_base, costs)
```

Train Results:  
```{r}
train_results_gbm$results
```

## Random Forest
```{r, include=FALSE}
h2o.loadGrid(
  "models/drf_grid_11"
)

drf_gridperf_2_byPrecision <- h2o.getGrid(grid_id = "drf_grid_11",
                                          sort_by = "precision",
                                          decreasing = TRUE)

drf_gridperf_2_byAUC <- h2o.getGrid(grid_id = "drf_grid_11",
                                    sort_by = "auc",
                                    decreasing = TRUE)

# Grab the top GBM model, chosen by validation AUC
model_drf_best <- h2o.getModel(drf_gridperf_2_byAUC@model_ids[[1]])
```

```{r, include=FALSE}
train_pred_df_drf_base <- get_prediction(model_drf_best, train_h2o)
train_results_drf <- get_results(train_pred_df_drf_base, costs)
opt_threshold_drf = optimize(get_cost_given_threshold, c(0.1, 0.9), truth_pred_df=train_pred_df_drf_base, costs=costs)

test_pred_df_drf_base_default <- get_prediction(model_drf_best, test_h2o)
test_pred_df_drf_base <- get_prediction_class_given_threshold(test_pred_df_drf_base_default, opt_threshold_drf$minimum)
test_results_drf <- get_results(test_pred_df_drf_base, costs)
```

Train Results:  
```{r}
train_results_drf$results
```

## Comparison
```{r}
model_comparison_df <- data.frame(
  models = c("baseline", "Logistic Regression", "GBM", "Random Forest"),
  train_credit_cost = c(baseline_train_cost, opt_threshold_glm$objective, opt_threshold_gbm$objective, opt_threshold_drf$objective),
  train_precision = c(baseline_train_precision, train_results_glm$results$byClass["Pos Pred Value"], train_results_gbm$results$byClass["Pos Pred Value"], train_results_drf$results$byClass["Pos Pred Value"]),
  test_credit_cost = c(baseline_test_cost, test_results_glm$cost, test_results_gbm$cost, test_results_drf$cost),
  test_precision = c(baseline_test_precision, test_results_glm$results$byClass["Pos Pred Value"], test_results_gbm$results$byClass["Pos Pred Value"], test_results_drf$results$byClass["Pos Pred Value"])
)

model_comparison_df
```

> Credit_cost and Pricision are in sync.

> train results are best for GBM. But its overfitting, i.e. variance is high, so not that great results on test.

> test results are best for Random Forest. It has less variance then GBM, but bias is higher.


It may seem like that GBM is a better model, but we still haven't seen the uncertainity (variance) in the results. Difference between train and test set results give some idea about it, but its better to see it on cross-validated results.

```{r}
temp <- model_gbm_best@model$cross_validation_metrics_summary
temp_1 <- tibble(metrics = row.names(temp)) %>% 
  bind_cols(temp %>% as_tibble())
datatable(temp_1[,1:3], rownames = FALSE, filter="none", options = list(scrollX=T, scrollY="200px", scrollCollapse=FALSE, paging=FALSE))
```

```{r}
temp <- model_drf_best@model$cross_validation_metrics_summary
temp_1 <- data.frame(metrics = row.names(temp)) %>% 
  bind_cols(temp %>% as_tibble())
datatable(temp_1[,1:3], rownames = FALSE, filter="none", options = list(scrollX=T, scrollY="200px", scrollCollapse=FALSE, paging=FALSE))
```

Not much difference here too, DRF seems only slightly better but that may change with fold assignment. For GBM, I did positive class upsample tuning but didn't tune other hyperparameters. And for DRF I did the exact opposite. So, both the models have a lot of scope of tuning, and I am not at a stage to pick the right model

# Important Features
We can see feature importance of either GBM or DRF, but DRF gives a better plot without breaking categorical features into its classes, so we will use DRF.

```{r}
h2o.varimp_plot(model_drf_best)
```

> Topp-3 features are "checking_account_status", "duration_in_months", and "credit_amount"

# Profiling of best credit-worthy person
To profile a 'Good' credit worthy person as per the model, let's explore the relationship of top predictors with the predicted class for the DRF model.


```{r, include=FALSE}
train_pred_df_drf_default <- get_prediction(model_drf_best, train_h2o)
train_pred_df_drf_opt <- get_prediction_class_given_threshold(train_pred_df_drf_default, opt_threshold_drf$minimum)

test_pred_df_drf_default <- get_prediction(model_drf_best, test_h2o)
test_pred_df_drf_opt <- get_prediction_class_given_threshold(test_pred_df_drf_default, opt_threshold_drf$minimum)

all_pred_df_drf_opt <- bind_rows(train_pred_df_drf_opt, test_pred_df_drf_opt)
```

```{r, message=FALSE}
plot_list = plot_target_cat_feature_cat(all_pred_df_drf_opt, "checking_account_status", "is_credit_worthy_pred_class", "Good", "Bad")

ggplotly(
  plot_list[[1]]
)
```

```{r, message=FALSE}
plot_list = plot_target_cat_feature_cont(all_pred_df_drf_opt, "duration_in_months", "is_credit_worthy_pred_class")

ggplotly(
  plot_list[[1]]
)
```

```{r, message=FALSE}
plot_list = plot_target_cat_feature_cont(all_pred_df_drf_opt, "credit_amount", "is_credit_worthy_pred_class")

ggplotly(
  plot_list[[1]]
)
```

```{r, message=FALSE}
plot_list = plot_target_cat_feature_cat(all_pred_df_drf_opt, "credit_history", "is_credit_worthy_pred_class", "Good", "Bad")

ggplotly(
  plot_list[[1]]
)
```

```{r, message=FALSE}
plot_list = plot_target_cat_feature_cat(all_pred_df_drf_opt, "purpose", "is_credit_worthy_pred_class", "Good", "Bad")

ggplotly(
  plot_list[[1]]
)
```

So, the best credit worthy person would have a following profile:  
- checking_account_status is "A14" i.e. no checking account  
- duration_in_months is less than 12 month i.e. a year  
- credit_amount is less than 2k  
- credit_history is "A34" i.e. critical account/other existing credits  
- Purpose is A43 i.e. radio/television  

This seems slightly unintuitive, but I will have to go into model explainibility to get better insights, and currently the time is short for that

# Things to do in future  
- EDA driven Feature Engineering  
- EDA driven Feature Selection  
- Better tuning  
  - with cross validation on credit_cost or Precision  
  - Bayesian Optimiation  
- Model Explainibility  